#endif
if (pte) {
clear_highpage(pte);
- __make_page_readonly(phys_to_virt(page_to_phys(pte)));
+ __make_page_readonly(phys_to_virt(page_to_pseudophys(pte)));
/* XXXcl highmem */
}
return pte;
ring_req->nr_segments = 0;
rq_for_each_bio(bio, req) {
bio_for_each_segment(bvec, bio, idx) {
- buffer_ma =
- phys_to_machine(page_to_phys(bvec->bv_page));
+ buffer_ma = page_to_phys(bvec->bv_page);
if (unlikely((buffer_ma & ((1<<9)-1)) != 0))
BUG();
for (i = 0; i < nents; i++ ) {
BUG_ON(!sg[i].page);
- sg[i].dma_address = page_to_machine(sg[i].page) + sg[i].offset;
+ sg[i].dma_address = page_to_phys(sg[i].page) + sg[i].offset;
}
flush_write_buffers();
size_t size, enum dma_data_direction direction)
{
BUG_ON(direction == DMA_NONE);
- return page_to_machine(page) + offset;
+ return page_to_phys(page) + offset;
}
static inline void
/*
* Change "struct page" to physical address.
*/
-#define page_to_phys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
-#define page_to_machine(page) (phys_to_machine(page_to_phys(page)))
+#define page_to_pseudophys(page) ((dma_addr_t)page_to_pfn(page) << PAGE_SHIFT)
+#define page_to_phys(page) (phys_to_machine(page_to_pseudophys(page)))
extern void * __ioremap(unsigned long offset, unsigned long size, unsigned long flags);